Ghibli creature 2
Customize
Thomas Kwa
*Ω321950
1
Some versions of the METR time horizon paper from alternate universes: Measuring AI Ability to Take Over Small Countries (idea by Caleb Parikh) Abstract: Many are worried that AI will take over the world, but extrapolation from existing benchmarks suffers from a large distributional shift that makes it difficult to forecast the date of world takeover. We rectify this by constructing a suite of 193 realistic, diverse countries with territory sizes from 0.44 to 17 million km^2. Taking over most countries requires acting over a long time horizon, with the exception of France. Over the last 6 years, the land area that AI can successfully take over with 50% success rate has increased from 0 to 0 km^2, doubling 0 times per year (95% CI 0.0-0.0 yearly doublings); extrapolation suggests that AI world takeover is unlikely to occur in the near future. To address concerns about the narrowness of our distribution, we also study AI ability to take over small planets and asteroids, and find similar trends. When Will Worrying About AI Be Automated? Abstract: Since 2019, the amount of time LW has spent worrying about AI has doubled every seven months, and now constitutes the primary bottleneck to AI safety research. Automation of worrying would be transformative to the research landscape, but worrying includes several complex behaviors, ranging from simple fretting to concern, anxiety, perseveration, and existential dread, and so is difficult to measure. We benchmark the ability of frontier AIs to worry about common topics like disease, romantic rejection, and job security, and find that current frontier models such as Claude 3.7 Sonnet already outperform top humans, especially in existential dread. If these results generalize to worrying about AI risk, AI systems will be capable of autonomously worrying about their own capabilities by the end of this year, allowing us to outsource all our AI concerns to the systems themselves. Estimating Time Since The Singularity Early work
Yonatan Cale
1470
1
Seems like Unicode officially added a "person being paperclipped" emoji: Here's how it looks in your browser: 🙂‍↕️ Whether they did this as a joke or to raise awareness of AI risk, I like it! Source: https://emojipedia.org/emoji-15.1
lc
940
7
My strong upvotes are now giving +1 and my regular upvotes give +2.
keltan
404
0
I feel a deep love and appreciation for this place, and the people who inhabit it.
RobertM
400
0
Pico-lightcone purchases are back up, now that we think we've ruled out any obvious remaining bugs.  (But do let us know if you buy any and don't get credited within a few minutes.)

Popular Comments

Recent Discussion

Epistemic status: This post aims at an ambitious target: improving intuitive understanding directly. The model for why this is worth trying is that I believe we are more bottlenecked by people having good intuitions guiding their research than, for example, by the ability of people to code and run evals. 

Quite a few ideas in AI safety implicitly use assumptions about individuality that ultimately derive from human experience. 

When we talk about AIs scheming, alignment faking or goal preservation, we imply there is something scheming or alignment faking or wanting to preserve its goals or escape the datacentre.

If the system in question were human, it would be quite clear what that individual system is. When you read about Reinhold Messner reaching the summit of Everest, you would be curious about...

(Other than the thoughts on the consequences of said idea) This idea largely seems like a rehash of https://www.lesswrong.com/posts/vJFdjigzmcXMhNTsx/simulators (and frankly, so does the three layer model, but that does go into more mechanistic territory and I think it complements simulator theory well)

This post was adapted from an internal doc I wrote at Wave.

Welcome to being a manager! Your time-management problem just got a lot harder.

As an IC, you can often get away with a very simple time-management strategy:

  1. Decide what your one most important thing is.
  2. Work on it until it’s done.
  3. GOTO 1

As a team lead, this isn’t going to work, because much more of your work is interrupt-driven or gets blocked for long periods of time. One-on-ones! Code reviews! Design reviews! Prioritization meetings! Project check-ins! All of these are subject to an external schedule rather than being the type of thing that you can push on in a single focused block until it’s done.

Being a team lead means three big changes for your time management:

  1. You no longer have a

...

A key step in the classic argument for AI doom is instrumental convergence: the idea that agents with many different goals will end up pursuing the same few subgoals, which includes things like "gain as much power as possible".

If it wasn't for instrumental convergence, you might think that only AIs with very specific goals would try to take over the world. But instrumental convergence says it's the other way around: only AIs with very specific goals will refrain from taking over the world.

For idealised pure consequentialists -- agents that have an outcome they want to bring about, and do whatever they think will cause it -- some version of instrumental convergence seems surely true[1].

But what if we get AIs that aren't pure consequentialists, for example because they're...

Something entirely new occurred around March 26th, 2025. Following the release of OpenAI’s 4o image generation, a specific aesthetic didn’t just trend—it swept across the virtual landscape like a tidal wave. Scroll through timelines, and nearly every image, every meme, every shared moment seemed spontaneously re-rendered in the unmistakable style of Studio Ghibli. This wasn’t just another filter; it felt like a collective, joyful migration into an alternate visual reality.

But why? Why this specific style? And what deeper cognitive or technological threshold did we just cross? The Ghiblification wave wasn’t mere novelty; it was, I propose, the first widely experienced instance of successful reality transfer: the mapping of our complex, nuanced reality into a fundamentally different, yet equally coherent and emotionally resonant, representational framework.

And Ghibli, it turns out, was...

Raemon
30

A thing that gave me creeping horror about the Ghiblification is that the I don't think the masses actually particularly understand Ghibli. And the result is an uneven simulacrum-mask that gives the impression of "rendered with love and care" without actually being so. 

The Ghibli aesthetic is actually historically pretty important to me, and in particular important as a counterbalanacing force against, among other things, what I expect to happen by default with AI. Some things I like about Ghibli:

  • The "cinematic lens" emphasizes a kind of "see everythi
... (read more)

Our community is not prepared for an AI crash. We're good at tracking new capability developments, but not as much the company financials. Currently, both OpenAI and Anthropic are losing $5 billion+ a year, while under threat of losing users to cheap LLMs.

A crash will weaken the labs. Funding-deprived and distracted, execs struggle to counter coordinated efforts to restrict their reckless actions. Journalists turn on tech darlings. Optimism makes way for mass outrage, for all the wasted money and reckless harms.

You may not think a crash is likely. But if it happens, we can turn the tide.

Preparing for a crash is our best bet.[1] But our community is poorly positioned to respond. Core people positioned themselves inside institutions – to advise on how to maybe make AI 'safe',...

17Vladimir_Nesov
The scale of training and R&D spending by AI companies can be reduced on short notice, while global inference buildout costs much more and needs years of use to pay for itself. So an AI slowdown mostly hurts clouds and makes compute cheap due to oversupply, which might be a wash for AI companies. Confusingly major AI companies are closely tied to cloud providers, but OpenAI is distancing itself from Microsoft, and Meta and xAI are not cloud providers, so wouldn't suffer as much. In any case the tech giants will survive, it's losing their favor that seems more likely to damage AI companies, making them no longer able to invest as much in R&D.
2Polar
There is a possibility of self-reinforcing negative cycle: models don't show rapid capabilities improvement -> investors halt pouring money into AI sector -> AI labs focus on cutting costs -> models don't show rapid capabilities improvement.
48Seth Herd
Huh? Yes we're unprepared to capitalize on a crash because how would we? This post doesn't say how one might do that. It seems you've got ideas but why write this if you weren't going to say what they are or what you want us to do or think about?
Remmelt
10

Yes, I get you don’t just want to read about the problem but a potential solution. 

The next post in this sequence will summarise the plan by those experienced organisers.

These organisers led one of the largest grassroots movements in recent history. That took years of coalition building, and so will building a new movement. 

So they want to communicate the plan clearly, without inviting misinterpretations down the line. I myself rushed writing on new plans before (when I nuanced a press release put out by a time-pressed colleague at Stop AI). That... (read more)

I've been running meetups since 2019 in Kitchener-Waterloo. These were rationalist-adjacent from 2019-2021 (examples here) and then explicitly rationalist from 2022 onwards.

Here's a low-effort/stream of consciousness rundown of some meetups I ran in Q1 2025. Sometime late last year, I resolved to develop my meetup posts in such a way that they're more plug-and-play-able by other organizers who are interested in running meetups on the same topics. Below you'll find links to said meetup posts (which generally have an intro, required and supplemental readings, and discussion questions for sparking conversation—all free to take), and brief notes on how they went and how they can go better. Which is to say, this post might be kind of boring for non-organizers.

The Old Year and the New

The first meetup of...

Czynski
10

If there are a lot of people for the very-low-context NY meetup, possibly at least one very-low-context meetup per quarter is worth doing, to see if that gets people in/back more?

To get the best posts emailed to you, create an account! (2-3 posts per week, selected by the LessWrong moderation team.)
Log In Reset Password
...or continue with

Given that 'Seize control of the capital markets and use that unlimited funding to...' is on the critical path for a number of superintelligence doom scenarios, this paper might be interesting to the safety community.

IOSCO is an international body for coordinating capital market regulators, so this represents their collective view on the challenges presented by AI tech in the near to medium term.

https://www.iosco.org/library/pubdocs/pdf/IOSCOPD788.pdf

I think rationalists should consider taking more showers.

As Eliezer Yudkowsky once said, boredom makes us human. The childhoods of exceptional people often include excessive boredom as a trait that helped cultivate their genius:

A common theme in the biographies is that the area of study which would eventually give them fame came to them almost like a wild hallucination induced by overdosing on boredom. They would be overcome by an obsession arising from within.

Unfortunately, most people don't like boredom, and we now have little metal boxes and big metal boxes filled with bright displays that help distract us all the time, but there is still an effective way to induce boredom in a modern population: showering.

When you shower (or bathe, that also works), you usually are cut off...

10Gordon Seidoh Worley
More reasons to shower: smelling nice is good.
130Aella
Strong disagree. This is an ineffective way to create boredom. Showers are overly stimulating, with horrible changes in temperature, the sensation of water assaulting you nonstop, and requiring laborious motions to do the bare minimum of scrubbing required to make society not mad at you. A much better way to be bored is to go on a walk outside or lift weights at the gym or listen to me talk about my data cleaning issues
dkl9
62

"Stimulating" here is not quite the opposite of "boring". Many minds are used to said temperature changes, water assaults, and laborious motions, such that they still stimulate, but are easily ignored, leaving much space for thoughts. Showers are boring by consistency, despite stimulation.

5bohaska
I guess this is another case of 'Universal' Human Experiences That Not Everyone Has

PDF version. berkeleygenomics.org. Twitter thread. (Bluesky copy.)

Summary

The world will soon use human germline genomic engineering technology. The benefits will be enormous: Our children will be long-lived, will have strong and diverse capacities, and will be halfway to the end of all illness.

To quickly bring about this world and make it a good one, it has to be a world that is beneficial, or at least acceptable, to a great majority of people. What laws would make this world beneficial to most, and acceptable to approximately all? We'll have to chew on this question ongoingly.

Genomic Liberty is a proposal for one overarching principle, among others, to guide public policy and legislation around germline engineering. It asserts:

Parents have the right to freely choose the genomes of their children.

If upheld,...

1River
I think the frames in which you are looking at this are just completely wrong. We aren't really talking about "decisions about an individuals' reproduction". We are talking about how a parent can treat their child. This is something that is already highly regulated by the state, CPS is a thing, and it is good that it is a thing. There may be debates to be had about whether CPS has gone too far on certain issues, but there is a core sort of evil that CPS exists to address, and that it is good for the state to address. And blinding your child is a very core paradigmatic example of that sort of evil. Whether you do it by genetic engineering or surgically or through some other means is entirely beside the point. Genetic engineering isn't special. It is just another technology. To take something that is obviously wrong and evil when done by other means, that everyone will agree the state should prevent when done by other means, and say that the state should allow it when done by genetic engineering, that strikes me as a major political threat to genetic engineering. We don't get genetic engineering to happen by creating special rules for it that permit monstrosities forbidden by any other means. We get genetic engineering by showing people that it is just another technology, and we can use it to do good and not evil, applying the same notions of good and evil that we would anywhere else. If a blind parent asked a surgeon to sever the optic nerve of of their newborn baby, and the surgeon did it, both the parents and the surgeon would go to jail for child abuse. Any normal person can see that a genetic engineer should be subject to the same ethical and legal constraints there as the surgeon. Arguing otherwise will endanger your purported goal of promoting this technology.   This notion of "erasing a type of person" also seems like exactly the wrong frame for this. When we cured smallpox, did we erase the type of person called "smallpox survivor"? When we feed a hungry pe
TsviBT
10

Whether you do it by genetic engineering or surgically or through some other means is entirely beside the point. Genetic engineering isn't special.

I'm not especially distinguishing the methods, I'm mainly distinguishing whether it's being done to a living person. See my comment upthread https://www.lesswrong.com/posts/rxcGvPrQsqoCHndwG/the-principle-of-genomic-liberty?commentId=qnafba5dx6gwoFX4a

We get genetic engineering by showing people that it is just another technology, and we can use it to do good and not evil, applying the same notions of good a

... (read more)

LessOnline 2025

Ticket prices increase in 1 day

Join our Festival of Blogging and Truthseeking from May 30 - Jun 1, Berkeley, CA